LOLA is an ultra-large-scale multilingual large model based on the sparse Mixture-of-Experts (MoE) Transformer architecture, supporting over 160 languages, with competitive advantages in natural language generation and understanding tasks.
Large Language Model
Transformers Other